首页> 外文OA文献 >Stochastic Subgradient Algorithms for Strongly Convex Optimization over Distributed Networks
【2h】

Stochastic Subgradient Algorithms for Strongly Convex Optimization over Distributed Networks

机译:用于强凸优化的随机次梯度算法   分布式网络

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

We study diffusion and consensus based optimization of a sum of unknownconvex objective functions over distributed networks. The only access to thesefunctions is through stochastic gradient oracles, each of which is onlyavailable at a different node, and a limited number of gradient oracle calls isallowed at each node. In this framework, we introduce a convex optimizationalgorithm based on the stochastic gradient descent (SGD) updates. Particularly,we use a carefully designed time-dependent weighted averaging of the SGDiterates, which yields a convergence rate of$O\left(\frac{N\sqrt{N}}{T}\right)$ after $T$ gradient updates for each node ona network of $N$ nodes. We then show that after $T$ gradient oracle calls, theaverage SGD iterate achieves a mean square deviation (MSD) of$O\left(\frac{\sqrt{N}}{T}\right)$. This rate of convergence is optimal as itmatches the performance lower bound up to constant terms. Similar to the SGDalgorithm, the computational complexity of the proposed algorithm also scaleslinearly with the dimensionality of the data. Furthermore, the communicationload of the proposed method is the same as the communication load of the SGDalgorithm. Thus, the proposed algorithm is highly efficient in terms ofcomplexity and communication load. We illustrate the merits of the algorithmwith respect to the state-of-art methods over benchmark real life data sets andwidely studied network topologies.
机译:我们研究分布式网络上未知凸目标函数之和的基于扩散和共识的优化。对这些功能的唯一访问是通过随机梯度预言,每个梯度预言仅在不同的节点上可用,并且每个节点都允许有限数量的梯度预言调用。在此框架中,我们基于随机梯度下降(SGD)更新引入了凸优化算法。特别是,我们使用经过精心设计的SGDiterates的时间相关加权平均,在$ T $梯度更新后,其收敛速度为$ O \ left(\ frac {N \ sqrt {N}} {T} \ right)$ $ N $节点网络上的每个节点。然后,我们显示在$ T $梯度oracle调用之后,平均SGD迭代获得了$ O \ left(\ frac {\ sqrt {N}} {T} \ right)$的均方差(MSD)。这种收敛速度是最佳的,因为它与性能下限匹配至常数项。与SGD算法相似,该算法的计算复杂度也随数据的维数线性增长。此外,所提出方法的通信负载与SGD算法的通信负载相同。因此,所提出的算法在复杂性和通信负载方面是高效的。我们针对基准基准现实数据集和广泛研究的网络拓扑,说明了相对于最新方法的算法的优点。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号